Your browser doesn't support javascript.
Show: 20 | 50 | 100
Results 1 - 12 de 12
Filter
1.
Scand J Public Health ; 51(5): 735-743, 2023 Jul.
Article in English | MEDLINE | ID: covidwho-2319143

ABSTRACT

BACKGROUND: The association between tobacco smoking and the risk of COVID-19 and its adverse outcomes is controversial, as studies reported contrasting findings. Bias due to misclassification of the exposure in the analyses of current versus non-current smoking could be a possible explanation because former smokers may have higher background risks of the disease due to co-morbidity. The aim of the study was to investigate the extent of this potential bias by separating non-, former, and current smokers when assessing the risk or prognosis of diseases. METHODS: We analysed data from 43,400 participants in the Stockholm Public Health Cohort, Sweden, with information on smoking obtained prior to the pandemic. We estimated the risk of COVID-19, hospital admissions and death for (a) former and current smokers relative to non-smokers, (b) current smokers relative to non-current smokers, that is, including former smokers; adjusting for potential confounders (aRR). RESULTS: The aRR of a COVID-19 diagnosis was elevated for former smokers compared with non-smokers (1.07; 95% confidence interval (CI) =1.00-1.15); including hospital admission with any COVID-19 diagnosis (aRR= 1.23; 95% CI = 1.03-1.48); or with COVID-19 as the main diagnosis (aRR=1.23, 95% CI= 1.01-1.49); and death within 30 days with COVID-19 as the main or a contributory cause (aRR=1.40; 95% CI=1.00-1.95). Current smoking was negatively associated with risk of COVID-19 (aRR=0.79; 95% CI=0.68-0.91). CONCLUSIONS: Separating non-smokers from former smokers when assessing the disease risk or prognosis is essential to avoid bias. However, the negative association between current smoking and the risk of COVID-19 could not be entirely explained by misclassification.


Subject(s)
COVID-19 , Smokers , Humans , Tobacco , Public Health , COVID-19 Testing , COVID-19/epidemiology
2.
Eur Heart J ; 43(35): 3312-3322, 2022 09 14.
Article in English | MEDLINE | ID: covidwho-2255633

ABSTRACT

This review will discuss the limitations of data collected by RCTs in relation to their applicability to daily life clinical management. It will then argue that these limitations are only partially overcome by modifications of RCT design and conduction (e.g. 'pragmatic trials') while being substantially attenuated by real-life-derived research, which can fill many gaps left by trial-collected evidence and have thus an important complementary value. The focus will be on the real-life research approach based on the retrospective analysis of the now widely available healthcare utilization databases (formerly known as administrative databases), which will be discussed in detail for their multiple advantages as well as challenges. Emphasis will be given to the potential of these databases to provide low-cost information over long periods on many different healthcare issues, drug therapies in particular, from the general population to clinically important subgroups, including (i) prognostic aspects of treatments implemented at the medical practice level via hospitalization and fatality data and (ii) medical practice-related phenomena such as low treatment adherence and therapeutic inertia (unsatisfactorily evaluated by RCTs). It will also be mentioned that thanks to the current availability of these data in electronic format, results can be obtained quickly, helping timely decisions under emergencies. The potential shortcomings of this approach (confounding by indication, misclassification, and selection bias) will also be discussed along with their possible minimization by suitable analytic means. Finally, examples of the contributions of studies on hypertension and other cardiovascular risk factors will be offered based on retrospective healthcare utilization databases that have provided information on real-life cardiovascular treatments unavailable via RCTs.


Subject(s)
Hypertension , Research Design , Antihypertensive Agents/therapeutic use , Databases, Factual , Humans , Hypertension/drug therapy , Retrospective Studies
3.
Am J Epidemiol ; 2022 Oct 04.
Article in English | MEDLINE | ID: covidwho-2239048

ABSTRACT

Along with age and race, sex has historically been a core stratification and control variable in epidemiological research. While in recent decades research guidelines and institutionalized requirements have incorporated an approach differentiating biological sex from social gender, neither sex nor gender is itself a unidimensional construct. The conflation of dimensions within and between sex and gender presents a validity issue wherein proxy measures are used for dimensions of interest, often without explicit acknowledgement or evaluation. Individual-level dimensions of sex and gender are outlined as a guide for epidemiologists. Two case studies are presented. The first demonstrates how unacknowledged use of a sex/gender proxy for a sexed dimension of interest (uterine status) resulted in decades of cancer research misestimating risks, racial disparities, and age trends. The second illustrates how a multidimensional sex and gender framework may be applied to strengthen research on COVID-19 incidence, diagnosis, morbidity and mortality. Considerations are outlined, including 1) addressing the match between measures and theory, and explicitly acknowledging and evaluating proxy use; 2) improving measurement across dimensions and social ecological levels; 3) incorporating multidimensionality into research objectives; 4) interpreting sex, gender, and their effects as biopsychosocial.

4.
Int J Epidemiol ; 2022 Dec 06.
Article in English | MEDLINE | ID: covidwho-2233305

ABSTRACT

BACKGROUND: Non-random selection of analytic subsamples could introduce selection bias in observational studies. We explored the potential presence and impact of selection in studies of SARS-CoV-2 infection and COVID-19 prognosis. METHODS: We tested the association of a broad range of characteristics with selection into COVID-19 analytic subsamples in the Avon Longitudinal Study of Parents and Children (ALSPAC) and UK Biobank (UKB). We then conducted empirical analyses and simulations to explore the potential presence, direction and magnitude of bias due to this selection (relative to our defined UK-based adult target populations) when estimating the association of body mass index (BMI) with SARS-CoV-2 infection and death-with-COVID-19. RESULTS: In both cohorts, a broad range of characteristics was related to selection, sometimes in opposite directions (e.g. more-educated people were more likely to have data on SARS-CoV-2 infection in ALSPAC, but less likely in UKB). Higher BMI was associated with higher odds of SARS-CoV-2 infection and death-with-COVID-19. We found non-negligible bias in many simulated scenarios. CONCLUSIONS: Analyses using COVID-19 self-reported or national registry data may be biased due to selection. The magnitude and direction of this bias depend on the outcome definition, the true effect of the risk factor and the assumed selection mechanism; these are likely to differ between studies with different target populations. Bias due to sample selection is a key concern in COVID-19 research based on national registry data, especially as countries end free mass testing. The framework we have used can be applied by other researchers assessing the extent to which their results may be biased for their research question of interest.

5.
25th Congress of the Portuguese Statistical Society, SPE 2021 ; 398:215-225, 2022.
Article in English | Scopus | ID: covidwho-2173616

ABSTRACT

The diagnosis of ME/CFS is problematic due to the absence of a disease specific biomarker. As such, it is conducted under uncertainty using symptom-based criteria and the exclusion of known diseases. The possibility of misdiagnosing patients reduces the power to detect new and previously identified factors that can be associated with the disease. To investigate this problem, we previously conducted a simulation study to estimate the power of case-control association studies as a function of the misdiagnosed rate. Here we extended this simulation study to the more general situation where there is also the possibility of having misclassification in a binary factor related to a previous exposure to a given infection. Given the suggested link between ME/CFS and past viral infections including SARS-CoV-2 (which causes COVID-19), we performed the simulation study in the specific context of serological testing of this new coronavirus using published data from Portuguese, Spanish and Iranian seroepidemiological studies. © 2022, The Author(s), under exclusive license to Springer Nature Switzerland AG.

6.
J Med Libr Assoc ; 109(4): 609-612, 2021 Oct 01.
Article in English | MEDLINE | ID: covidwho-1538713

ABSTRACT

OBJECTIVE: We recently showed that genderize.io is not a sufficiently powerful gender detection tool due to a large number of nonclassifications. In the present study, we aimed to assess whether the accuracy of inference by genderize.io can be improved by manipulating the first names in the database. METHODS: We used a database containing the first names, surnames, and gender of 6,131 physicians practicing in a multicultural country (Switzerland). We uploaded the original CSV file (file #1), the file obtained after removing all diacritic marks, such as accents and cedilla (file #2), and the file obtained after removing all diacritic marks and retaining only the first term of the compound first names (file #3). For each file, we computed three performance metrics: proportion of misclassifications (errorCodedWithoutNA), proportion of nonclassifications (naCoded), and proportion of misclassifications and nonclassifications (errorCoded). RESULTS: naCoded, which was high for file #1 (16.4%), was reduced after data manipulation (file #2: 11.7%, file #3: 0.4%). As the increase in the number of misclassifications was small, the overall performance of genderize.io (i.e., errorCoded) improved, especially for file #3 (file #1: 17.7%, file #2: 13.0%, and file #3: 2.3%). CONCLUSIONS: A relatively simple manipulation of the data improved the accuracy of gender inference by genderize.io. We recommend using genderize.io only with files that were modified in this way.


Subject(s)
Gender Identity , Names , Data Collection
7.
J Med Libr Assoc ; 109(3): 414-421, 2021 Jul 01.
Article in English | MEDLINE | ID: covidwho-1481112

ABSTRACT

OBJECTIVE: To evaluate the performance of gender detection tools that allow the uploading of files (e.g., Excel or CSV files) containing first names, are usable by researchers without advanced computer skills, and are at least partially free of charge. METHODS: The study was conducted using four physician datasets (total number of physicians: 6,131; 50.3% female) from Switzerland, a multilingual country. Four gender detection tools met the inclusion criteria: three partially free (Gender API, NamSor, and genderize.io) and one completely free (Wiki-Gendersort). For each tool, we recorded the number of correct classifications (i.e., correct gender assigned to a name), misclassifications (i.e., wrong gender assigned to a name), and nonclassifications (i.e., no gender assigned). We computed three metrics: the proportion of misclassifications excluding nonclassifications (errorCodedWithoutNA), the proportion of nonclassifications (naCoded), and the proportion of misclassifications and nonclassifications (errorCoded). RESULTS: The proportion of misclassifications was low for all four gender detection tools (errorCodedWithoutNA between 1.5 and 2.2%). By contrast, the proportion of unrecognized names (naCoded) varied: 0% for NamSor, 0.3% for Gender API, 4.5% for Wiki-Gendersort, and 16.4% for genderize.io. Using errorCoded, which penalizes both types of error equally, we obtained the following results: Gender API 1.8%, NamSor 2.0%, Wiki-Gendersort 6.6%, and genderize.io 17.7%. CONCLUSIONS: Gender API and NamSor were the most accurate tools. Genderize.io led to a high number of nonclassifications. Wiki-Gendersort may be a good compromise for researchers wishing to use a completely free tool. Other studies would be useful to evaluate the performance of these tools in other populations (e.g., Asian).


Subject(s)
Gender Identity , Information Storage and Retrieval , Databases, Bibliographic , Female , Humans , Male
8.
Am J Epidemiol ; 190(8): 1681-1688, 2021 08 01.
Article in English | MEDLINE | ID: covidwho-1337251

ABSTRACT

We evaluated whether randomly sampling and testing a set number of individuals for coronavirus disease 2019 (COVID-19) while adjusting for misclassification error captures the true prevalence. We also quantified the impact of misclassification error bias on publicly reported case data in Maryland. Using a stratified random sampling approach, 50,000 individuals were selected from a simulated Maryland population to estimate the prevalence of COVID-19. We examined the situation when the true prevalence is low (0.07%-2%), medium (2%-5%), and high (6%-10%). Bayesian models informed by published validity estimates were used to account for misclassification error when estimating COVID-19 prevalence. Adjustment for misclassification error captured the true prevalence 100% of the time, irrespective of the true prevalence level. When adjustment for misclassification error was not done, the results highly varied depending on the population's underlying true prevalence and the type of diagnostic test used. Generally, the prevalence estimates without adjustment for misclassification error worsened as the true prevalence level increased. Adjustment for misclassification error for publicly reported Maryland data led to a minimal but not significant increase in the estimated average daily cases. Random sampling and testing of COVID-19 are needed with adjustment for misclassification error to improve COVID-19 prevalence estimates.


Subject(s)
COVID-19 Testing/statistics & numerical data , COVID-19/epidemiology , Decision Support Techniques , Statistics as Topic/methods , Bayes Theorem , COVID-19/classification , Humans , Maryland/epidemiology , Prevalence , SARS-CoV-2 , Selection Bias
9.
Clin Infect Dis ; 74(2): 368-370, 2022 01 29.
Article in English | MEDLINE | ID: covidwho-1227648
10.
Eur J Epidemiol ; 36(2): 179-196, 2021 Feb.
Article in English | MEDLINE | ID: covidwho-1103484

ABSTRACT

In response to the coronavirus disease (COVID-19) pandemic, public health scientists have produced a large and rapidly expanding body of literature that aims to answer critical questions, such as the proportion of the population in a geographic area that has been infected; the transmissibility of the virus and factors associated with high infectiousness or susceptibility to infection; which groups are the most at risk of infection, morbidity and mortality; and the degree to which antibodies confer protection to re-infection. Observational studies are subject to a number of different biases, including confounding, selection bias, and measurement error, that may threaten their validity or influence the interpretation of their results. To assist in the critical evaluation of a vast body of literature and contribute to future study design, we outline and propose solutions to biases that can occur across different categories of observational studies of COVID-19. We consider potential biases that could occur in five categories of studies: (1) cross-sectional seroprevalence, (2) longitudinal seroprotection, (3) risk factor studies to inform interventions, (4) studies to estimate the secondary attack rate, and (5) studies that use secondary attack rates to make inferences about infectiousness and susceptibility.


Subject(s)
COVID-19/epidemiology , Research Design , Bias , Humans , Reproducibility of Results , SARS-CoV-2 , Seroepidemiologic Studies
11.
Spat Spatiotemporal Epidemiol ; 36: 100401, 2021 02.
Article in English | MEDLINE | ID: covidwho-1014822

ABSTRACT

Surveillance data obtained by public health agencies for COVID-19 are likely inaccurate due to undercounting and misdiagnosing. Using a Bayesian approach, we sought to reduce bias in the estimates of prevalence of COVID-19 in Philadelphia, PA at the ZIP code level. After evaluating various modeling approaches in a simulation study, we estimated true prevalence by ZIP code with and without conditioning on an area deprivation index (ADI). As of June 10, 2020, in Philadelphia, the observed citywide period prevalence was 1.5%. After accounting for bias in the surveillance data, the median posterior citywide true prevalence was 2.3% when accounting for ADI and 2.1% when not. Overall the median posterior surveillance sensitivity and specificity from the models were similar, about 60% and more than 99%, respectively. Surveillance of COVID-19 in Philadelphia tends to understate discrepancies in burden for the more affected areas, potentially misinforming mitigation priorities.


Subject(s)
Bayes Theorem , COVID-19/epidemiology , Population Surveillance , Spatial Analysis , Bias , Humans , Philadelphia/epidemiology , Prevalence , SARS-CoV-2 , Sensitivity and Specificity
12.
Can J Public Health ; 111(3): 397-400, 2020 06.
Article in English | MEDLINE | ID: covidwho-1005629

ABSTRACT

During an epidemic with a new virus, we depend on modelling to plan the response: but how good are the data? The aim of our work was to better understand the impact of misclassification errors in identification of true cases of COVID-19 on epidemic curves. Data originated from Alberta, Canada (available on 28 May 2020). There is presently no information of sensitivity (Sn) and specificity (Sp) of laboratory tests used in Canada for the causal agent for COVID-19. Therefore, we examined best attainable performance in other jurisdictions and similar viruses. This suggested perfect Sp and Sn 60-95%. We used these values to re-calculate epidemic curves to visualize the potential bias due to imperfect testing. If the sensitivity improved, the observed and adjusted epidemic curves likely fall within 95% confidence intervals of the observed counts. However, bias in shape and peak of the epidemic curves can be pronounced, if sensitivity either degrades or remains poor in the 60-70% range. These issues are minor early in the epidemic, but hundreds of undiagnosed cases are likely later on. It is therefore hazardous to judge progress of the epidemic based on observed epidemic curves unless quality of testing is better understood.


Subject(s)
Clinical Laboratory Techniques/standards , Coronavirus Infections/diagnosis , Coronavirus Infections/epidemiology , Epidemics , Pneumonia, Viral/diagnosis , Pneumonia, Viral/epidemiology , Bias , COVID-19 , COVID-19 Testing , Canada/epidemiology , Humans , Pandemics , Probability , Sensitivity and Specificity
SELECTION OF CITATIONS
SEARCH DETAIL